127 research outputs found
Optimal mobility-aware admission control in content delivery networks
This paper addresses the problem of mobility management in Content Delivery Networks (CDN). We introduce a CDN architecture where admission control is performed at mobility aware access routers. We formulate a Markov Modulated Poisson Decision Process for access control that captures the bursty nature of data and packetized traffic together with the heterogeneity of multimedia services. The optimization of performance parameters, like the blocking probabilities and the overall utilization, is conducted and the structural properties of the optimal solutions are also studied. Heuristics are proposed to encompass the computational difficulties of the optimal solution when several classes of multimedia traffic are considered
Self-* overload control for distributed web systems
Unexpected increases in demand and most of all flash crowds are considered
the bane of every web application as they may cause intolerable delays or even
service unavailability. Proper quality of service policies must guarantee rapid
reactivity and responsiveness even in such critical situations. Previous
solutions fail to meet common performance requirements when the system has to
face sudden and unpredictable surges of traffic. Indeed they often rely on a
proper setting of key parameters which requires laborious manual tuning,
preventing a fast adaptation of the control policies. We contribute an original
Self-* Overload Control (SOC) policy. This allows the system to self-configure
a dynamic constraint on the rate of admitted sessions in order to respect
service level agreements and maximize the resource utilization at the same
time. Our policy does not require any prior information on the incoming traffic
or manual configuration of key parameters. We ran extensive simulations under a
wide range of operating conditions, showing that SOC rapidly adapts to time
varying traffic and self-optimizes the resource utilization. It admits as many
new sessions as possible in observance of the agreements, even under intense
workload variations. We compared our algorithm to previously proposed
approaches highlighting a more stable behavior and a better performance.Comment: The full version of this paper, titled "Self-* through self-learning:
overload control for distributed web systems", has been published on Computer
Networks, Elsevier. The simulator used for the evaluation of the proposed
algorithm is available for download at the address:
http://www.dsi.uniroma1.it/~novella/qos_web
Fundamental limits of failure identifiability by Boolean Network Tomography
Boolean network tomography is a powerful tool to infer the state (working/failed) of individual nodes from path-level measurements obtained by egde-nodes. We consider the problem of optimizing the capability of identifying network failures through the design of monitoring schemes. Finding an optimal solution is NP-hard and a large body of work has been devoted to heuristic approaches providing lower bounds. Unlike previous works, we provide upper bounds on the maximum number of identifiable nodes, given the number of monitoring paths and different constraints on the network topology, the routing scheme, and the maximum path length. The proposed upper bounds represent a fundamental limit on the identifiability of failures via Boolean network tomography. This analysis provides insights on how to design topologies and related monitoring schemes to achieve the maximum identifiability under various network settings. Through analysis and experiments we demonstrate the tightness of the bounds and efficacy of the design insights for engineered as well as real network
Self-Adaptive resource allocation for event monitoring with uncertainty in Sensor Networks
Event monitoring is an important application of sensor networks. Multiple parties, with different surveillance targets, can share the same network, with limited sensing resources, to monitor their events of interest simultaneously.
Such a system achieves profit by allocating sensing resources to missions to collect event related information (e.g., videos, photos, electromagnetic signals). We address the problem of dynamically
assigning resources to missions so as to achieve maximum profit with uncertainty in event occurrence. We consider timevarying resource demands and profits, and multiple concurrent surveillance missions. We model each mission as a sequence of monitoring attempts, each being allocated with a certain amount of resources, on a specific set of events that occurs as a
Markov process. We propose a Self-Adaptive Resource Allocation algorithm (SARA) to adaptively and efficiently allocate resources according to the results of previous observations. By means of simulations we compare SARA to previous solutions and show SARA’s potential in finding higher profit in both static and dynamic scenarios
Power adjustment and scheduling in OFDMA femtocell networks
Densely-deployed femtocell networks are used to enhance wireless coverage in public spaces like office buildings, subways, and academic buildings. These networks can increase throughput for users, but edge users can suffer from co-channel interference, leading to service outages. This paper introduces a distributed algorithm for network configuration, called Radius Reduction and Scheduling (RRS), to improve the performance and fairness of the network. RRS determines cell sizes using a Voronoi-Laguerre framework, then schedules users using a scheduling algorithm that includes vacancy requests to increase fairness in dense femtocell networks. We prove that our algorithm always terminate in a finite time, producing a configuration that guarantees user or area coverage. Simulation results show a decrease in outage probability of up to 50%, as well as an increase in Jain's fairness index of almost 200%
Call Admission Control : Solution of a General Decision Model with State Related Hand-off Rate
This paper studies call admission policies for access control in cellular networks by means of a Markov Decision
Process (MDP). This approach allows us to study a wide class of policies, including well known pure stationary as well
as randomized policies, in a way that explicitly incorporates the dependency between the hand-off rate and the system
state, assuming that the hand-off rate arriving to a cell is proportional to the occupancy level of the adjacent cells.
In particular, we propose and analyze a nonpreemptive prioritization scheme, we term the cutoff priority policy. This
policy consists of reserving a number of channels for the high priority requests stream. Using our analytical approach,
we prove the proposed scheme to be optimal within the analyzed class
Progressive damage assessment and network recovery after massive failures
After a massive scale failure, the assessment of damages to communication networks requires local interventions and remote monitoring. While previous works on network recovery require complete knowledge of damage extent, we address the problem of damage assessment and critical service restoration in a joint manner. We propose a polynomial algorithm called Centrality based Damage Assessment and Recovery (CeDAR) which performs a joint activity of failure monitoring and restoration of network components. CeDAR works under limited availability of recovery resources and optimizes service recovery over time. We modified two existing approaches to the problem of network recovery to make them also able to exploit incremental knowledge of the failure extent. Through simulations we show that CeDAR outperforms the previous approaches in terms of recovery resource utilization and accumulative flow over time of the critical service
Energy-Efficient selective activation in Femtocell Networks
Provisioning the capacity of wireless networks is difficult when peak load is significantly higher than average load, for example, in public spaces like airports or train stations. Service providers can use femtocells and small cells to increase local capacity, but deploying enough femtocells to serve peak loads requires a large number of femtocells that will remain idle most of the time, which wastes a significant amount of power.
To reduce the energy consumption of over-provisioned femtocell networks, we formulate a femtocell selective activation problem, which we formalize as an integer nonlinear optimization problem. Then we introduce GREENFEMTO, a distributed femtocell selective activation algorithm that deactivates idle femtocells to
save power and activates them on-the-fly as the number of users increases. We prove that GREENFEMTO converges to a locally Pareto optimal solution and demonstrate its performance using extensive simulations of an LTE wireless system. Overall, we find that GREENFEMTO requires up to 55% fewer femtocells to serve a given user load, relative to an existing femtocell power-saving procedure, and comes within 15% of a globally optimal solution
On the vulnerabilities of voronoi-based approaches to mobile sensor deployment
Mobile sensor networks are the most promising solution to cover an Area of Interest (AoI) in safety critical scenarios. Mobile devices can coordinate with each other according to a distributed deployment algorithm, without resorting to human supervision for device positioning and network configuration. In this paper, we focus on the vulnerabilities of the deployment algorithms based on Voronoi diagrams to coordinate mobile sensors and guide their movements. We give a geometric characterization of possible attack configurations, proving that a simple attack consisting of a barrier of few compromised sensors can severely reduce network coverage. On the basis of the above characterization, we propose two new secure deployment algorithms, named SecureVor and Secure Swap Deployment (SSD). These algorithms allow a sensor to detect compromised nodes by analyzing their movements, under different and complementary operative settings. We show that the proposed algorithms are effective in defeating a barrier attack, and both have guaranteed termination. We perform extensive simulations to study the performance of the two algorithms and compare them with the original approach. Results show that SecureVor and SSD have better robustness and flexibility and excellent coverage capabilities and deployment time, even in the presence of an attac
Network recovery after massive failures
This paper addresses the problem of efficiently restoring sufficient resources in a communications network to support the demand of mission critical services after a large scale disruption. We give a formulation of the problem as an MILP and show that it is NP-hard. We propose a polynomial time heuristic, called Iterative Split and Prune (ISP) that decomposes the original problem recursively into smaller problems, until it determines the set of network components to be restored. We performed extensive simulations by varying the topologies, the demand intensity, the number of critical services, and the disruption model. Compared to several greedy approaches ISP performs better in terms of number of repaired components, and does not result in any demand loss. It performs very close to the optimal when the demand is low with respect to the supply network capacities, thanks to the ability of the algorithm to maximize sharing of repaired resources
- …